Goto

Collaborating Authors

 multi-modal dependency tree


Multi-modal Dependency Tree for Video Captioning

Neural Information Processing Systems

Generating fluent and relevant language to describe visual content is critical for the video captioning task. Many existing methods generate captions using sequence models that predict words in a left-to-right order. In this paper, we investigate a graph-structured model for caption generation by explicitly modeling the hierarchical structure in the sentences to further improve the fluency and relevance of sentences. To this end, we propose a novel video captioning method that generates a sentence by first constructing a multi-modal dependency tree and then traversing the constructed tree, where the syntactic structure and semantic relationship in the sentence are represented by the tree topology. To take full advantage of the information from both vision and language, both the visual and textual representation features are encoded into each tree node. Different from existing dependency parsing methods that generate uni-modal dependency trees for language understanding, our method construct s multi-modal dependency trees for language generation of images and videos. We also propose a tree-structured reinforcement learning algorithm to effectively optimize the captioning model where a novel reward is designed by evaluating the semantic consistency between the generated sub-tree and the ground-truth tree. Extensive experiments on several video captioning datasets demonstrate the effectiveness of the proposed method.


Supplementary Material for Multi-modal Dependency Tree for Video Captioning

Neural Information Processing Systems

The evaluation results on the Charades Captions dataset are shown in Table 2. Figure 1: Qualitative results of the generated tree structure and sentences on the MSR-VTT dataset. Table 1: The average sentence lengths of ground-truth captions and the captions generated by "w/o Lower average edit distance is better. In this section, we illustrate more details of human evaluation. We recruited 10 annotators to carry out the human evaluation process. The user interface for human evaluation is shown in Figure 4. To ensure Do the main claims made in the abstract and introduction accurately reflect the paper's Did you discuss any potential negative societal impacts of your work?



Multi-modal Dependency Tree for Video Captioning

Neural Information Processing Systems

Generating fluent and relevant language to describe visual content is critical for the video captioning task. Many existing methods generate captions using sequence models that predict words in a left-to-right order. In this paper, we investigate a graph-structured model for caption generation by explicitly modeling the hierarchical structure in the sentences to further improve the fluency and relevance of sentences. To this end, we propose a novel video captioning method that generates a sentence by first constructing a multi-modal dependency tree and then traversing the constructed tree, where the syntactic structure and semantic relationship in the sentence are represented by the tree topology. To take full advantage of the information from both vision and language, both the visual and textual representation features are encoded into each tree node.